Robust Learning with FeatureBoost

نویسندگان

  • Joseph O'Sullivan
  • John Langford
  • Rich Caruana
  • Avrim Blum
چکیده

PAC learning typically assumes that the training and test sets are drawn from the same distributions. This assumption often is violated in practice. How can machine learning algorithms be biased to output hypotheses that are robust to alterations in the test dis-tribution? We create a framework for learning in environments where the test and training distributions diier because features in the test sets are likely to be missing or corrupted. Motivated by this framework, we present a new meta-learning algorithm, Fea-tureBoost. We demonstrate FeatureBoost on three learning problems using backprop nets, k-nearest neighbor, and decision trees. 1 Motivation Consider a backprop net learning to steer a car in Pomerleau's ALVINN system 7]. The principle internal features learned by ALVINN nets detect the left and right edges of the road. ALVINN nets do not learn internal features that detect road centerlines. This creates a problem when the left or right edges of the road are obstructed by passing vehicles, or are missing as on bridges. Yet human steering is remarkably robust to the loss of these features because human drivers can fall back on a number of alternate features as diierent subsets of road features come in and out of view. Backprop nets do learn to steer better if they are forced to learn to recognize centerliness4]. Why do ALVINN nets not automatically learn to use a variety of road features (such as centerlines) when learning to steer? A related problem arises in pneumonia risk predictionn8]. Here there are a number of basic inputs available for patients before they enter the hospital (e.g., age, gender, blood pressure), as well as a number of lab features that become available after patients enter the hospital (e.g., RBC counts, oxygenation, Albumin). Models trained to predict risk from both the pre and in-hospital features are more accurate than models trained from only the pre-hospital inputs, but perform poorly on patients not yet admitted to the hospital (marginalizing over the missing in-hospital features). Models trained to predict risk from only the pre-hospital inputs are more accurate on patients not yet admitted to the hospital than marginalized models trained on all the features. If we can train models to make good predictions from only the pre-hospital inputs, why don't models trained with all the features work this well when some of the features are missing?

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

FeatureBoost: A Meta-Learning Algorithm that Improves Model Robustness

Most machine learning algorithms are lazy: they extract from the training set the minimum information needed to predict its labels. Unfortunately, this often leads to models that are not robust when features are removed or obscured in future test data. For example, a backprop net trained to steer a car typically learns to recognize the edges of the road, but does not learn to recognize other fe...

متن کامل

Simulation of Scour Pattern Around Cross-Vane Structures Using Outlier Robust Extreme Learning Machine

In this research, the scour hole depth at the downstream of cross-vane structures with different shapes (i.e., J, I, U, and W) was simulated utilizing a modern artificial intelligence method entitled "Outlier Robust Extreme Learning Machine (ORELM)". The observational data were divided into two groups: training (70%) and test (30%). Then, using the input parameters including the ratio of the st...

متن کامل

Perfect Tracking of Supercavitating Non-minimum Phase Vehicles Using a New Robust and Adaptive Parameter-optimal Iterative Learning Control

In this manuscript, a new method is proposed to provide a perfect tracking of the supercavitation system based on a new two-state model. The tracking of the pitch rate and angle of attack for fin and cavitator input is of the aim. The pitch rate of the supercavitation with respect to fin angle is found as a non-minimum phase behavior. This effect reduces the speed of command pitch rate. Control...

متن کامل

Alleviating the Small-Signal Oscillations of the SMIB Power System with the TLBO–FPSS and SSSC Robust Controller

Power systems are subjected to small–signal oscillations that can be caused by sudden change in the value of large loads. To avoid the dangers of these oscillations, the Power System Stabilizers (PSSs) are used. When the PSSs can not be effective enough, installation of the Thyristor–based compensators to increase the oscillations damping is a suitable method. In this paper, a Static Synchronou...

متن کامل

An Effective Approach for Robust Metric Learning in the Presence of Label Noise

Many algorithms in machine learning, pattern recognition, and data mining are based on a similarity/distance measure. For example, the kNN classifier and clustering algorithms such as k-means require a similarity/distance function. Also, in Content-Based Information Retrieval (CBIR) systems, we need to rank the retrieved objects based on the similarity to the query. As generic measures such as ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2007